Least relative entropy for voiced/unvoiced speech classification
نویسندگان
چکیده
The aim of this work is to develop ajlexible and eficient approach to the classifcation of the ratio of voiced to unvoiced excitation sources in continuous speech. To achieve this aim we adopt a probabilistic neural network approach. This is accomplished by designing a multi layer perceptron classifer trained by steepest descent minimization of the Least Relative Entropy W) cost function. By using the LRE cost function we can directly ou@ut the ratio, as aprobabiliv, of excitation source, voiced to unvoiced, for a given speech segment. These output probabilities can then be used directly in other applications, such as low bit rate coders.
منابع مشابه
Denoising Of Speech Signal By Classification Into Voiced, Unvoiced And Silence Region
In this paper, a speech enhancement method based on the classification of voiced, unvoiced and silence regions and using stationary wavelet transform is presented. To prevent the quality of degradation of speech during the denoising process, speech is first classified into voiced, unvoiced and silence regions. An experimentally verified criterion based on the short time energy process has been ...
متن کاملPerceptual entropy rate estimates for the phonemes of American English
We estimated the perceptual entropy rate of the phonemes of American English and found that the upper limit of the perceptual entropy of voiced phonemes is approximately 1.4 bit/sample, whereas the perceptual entropy of unvoiced phonemes is approximately 0.9 bit/sample. Results indicate that a simple voiced/unvoiced classi cation is suboptimal when trying to minimize bit rate. We used two di er...
متن کاملA Variable Rate Speech Codec Using Vus Classification
Voiced speech is highly correlated and must be reconstructed accurately in order to sound correct. Unvoiced speech on the other hand is noise like in nature. It can be approximated by white noise coloured by the vocal tract filter. Because of this lack of structure in unvoiced speech sounds, the excitation signal does not have to reproduce the speech signal as accurately as for voiced sounds. T...
متن کاملUniform concatenative excitation model for synthesising speech without voiced/unvoiced classification
In general, speech synthesis using the source-filter model of speech production requires the classification of speech into two classes (voiced and unvoiced) which is prone to errors. For voiced speech, the input of the synthesis filter is an approximately periodic excitation, whereas it is a noise signal for unvoiced. This paper proposes an excitation model which can be used to synthesise both ...
متن کاملVoiced - Unvoiced - Silence Classification via Hierarchical Dual Geometry Analysis
The need for a reliable discrimination among voiced, unvoiced and silence frames arises in a wide variety of speech processing applications. In this paper, we propose an unsupervised algorithm for voiced-unvoiced-silence classification based on a time-frequency representation of the measured signal, which is viewed as a data matrix. The proposed algorithm relies on a hierarchical dual geometry ...
متن کامل